Calibrated Fairness in Bandits

نویسندگان

  • Yang Liu
  • Goran Radanovic
  • Christos Dimitrakakis
  • Debmalya Mandal
  • David C. Parkes
چکیده

We study fairness within the stochastic,multi-armed bandit (MAB) decision making framework. We adapt the fairness framework of “treating similar individuals similarly” [5] to this seŠing. Here, an ‘individual’ corresponds to an arm and two arms are ‘similar’ if they have a similar quality distribution. First, we adopt a smoothness constraint that if two arms have a similar quality distribution then the probability of selecting each arm should be similar. In addition, we de€ne the fairness regret, which corresponds to the degree to which an algorithm is not calibrated, where perfect calibration requires that the probability of selecting an arm is equal to the probability with which the arm has the best quality realization. We show that a variation on Œompson sampling satis€es smooth fairness for total variation distance, and give an Õ((kT )2/3) bound on fairness regret. Œis complements prior work [12], which protects an on-average beŠer arm from being less favored. We also explain how to extend our algorithm to the dueling bandit seŠing. ACM Reference format: Yang Liu, Goran Radanovic, Christos Dimitrakakis, DebmalyaMandal, andDavid C. Parkes. 2017. Calibrated Fairness in Bandits. In Proceedings of FAT-ML, Calibrated Fairness in Bandits, September 2017 (FAT-ML17), 7 pages. DOI: 10.1145/nnnnnnn.nnnnnnn

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Better Fair Algorithms for Contextual Bandits ∗ Matthew

We study fairness in the linear bandit setting. Starting from the notion of meritocratic fairness introduced in Joseph et al. [11], we introduce a sufficiently more general model in which meritocratic fairness can be imposed and satisfied. We then perform a more fine-grained analysis which achieves better performance guarantees in this more general model. Our work therefore studies fairness for...

متن کامل

Fairness in Learning: Classic and Contextual Bandits

We introduce the study of fairness in multi-armed bandit problems. Our fairness definition demands that, given a pool of applicants, a worse applicant is never favored over a better one, despite a learning algorithm’s uncertainty over the true payoffs. In the classic stochastic bandits problem we provide a provably fair algorithm based on “chained” confidence intervals, and prove a cumulative r...

متن کامل

Meritocratic Fairness for Infinite and Contextual Bandits

We study fairness in linear bandit problems. Starting from the notion of meritocratic fairness introduced in Joseph et al. (2016), we carry out a more refined analysis of a more general problem, achieving better performance guarantees with fewer modelling assumptions on the number and structure of available choices as well as the number selected. We also analyze the previously-unstudied questio...

متن کامل

Fair Algorithms for Infinite Contextual Bandits

We study fairness in infinite linear bandit problems. Starting from the notion of meritocratic fairness introduced in Joseph et al. [9], we expand their notion of fairness for infinite action spaces and provide an algorithm that obtains a sublinear but instance-dependent regret guarantee. We then show that this instance dependence is a necessary cost of our fairness definition with a matching l...

متن کامل

Collaborative Spatial Reuse in Wireless Networks via Selfish Multi-Armed Bandits

Next-generation wireless deployments are characterized by being dense and uncoordinated, which often leads to inefficient use of resources and poor performance. To solve this, we envision the utilization of completely decentralized mechanisms that enhance Spatial Reuse (SR). In particular, we concentrate in Reinforcement Learning (RL), and more specifically, in Multi-Armed Bandits (MABs), to al...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1707.01875  شماره 

صفحات  -

تاریخ انتشار 2017